Continuous multi-task Bayesian Optimisation with correlation
نویسندگان
چکیده
منابع مشابه
Multi-fidelity Bayesian Optimisation with Continuous Approximations
Bandit methods for black-box optimisation, such as Bayesian optimisation, are used in a variety of applications including hyper-parameter tuning and experiment design. Recently, multifidelity methods have garnered considerable attention since function evaluations have become increasingly expensive in such applications. Multifidelity methods use cheap approximations to the function of interest t...
متن کاملBayesian Optimisation with Continuous Approximations
To present the theoretical results for GP-UCB, we begin by defining the Maximum Information Gain (MIG) which characterises the statistical difficulty of GP bandits. Definition 2. (Maximum Information Gain (Srinivas et al., 2010)) Let f ∼ GP(0, φX ). Consider any A ⊂ R and let A′ = {x1, . . . , xn} ⊂ A be a finite subset. Let fA′ , A′ ∈ R such that (fA′)i = f(xi) and ( A′)i ∼ N (0, η). Let yA′ =...
متن کاملMulti-Task Bayesian Optimization
Bayesian optimization has recently been proposed as a framework for automatically tuning the hyperparameters of machine learning models and has been shown to yield state-of-the-art performance with impressive ease and efficiency. In this paper, we explore whether it is possible to transfer the knowledge gained from previous optimizations to new tasks in order to find optimal hyperparameter sett...
متن کاملConcept Classification with Bayesian Multi-task Learning
Multivariate analysis allows decoding of single trial data in individual subjects. Since different models are obtained for each subject it becomes hard to perform an analysis on the group level. We introduce a new algorithm for Bayesian multi-task learning which imposes a coupling between single-subject models. Using the CMU fMRI dataset it is shown that the algorithm can be used for concept cl...
متن کاملSparse Bayesian Multi-Task Learning
We propose a new sparse Bayesian model for multi-task regression and classification. The model is able to capture correlations between tasks, or more specifically a low-rank approximation of the covariance matrix, while being sparse in the features. We introduce a general family of group sparsity inducing priors based on matrix-variate Gaussian scale mixtures. We show the amount of sparsity can...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: European Journal of Operational Research
سال: 2018
ISSN: 0377-2217
DOI: 10.1016/j.ejor.2018.03.017